knowledge graph completion
- Research Report > Experimental Study (0.93)
- Workflow (0.67)
- North America > United States > Illinois > Cook County > Chicago (0.04)
- North America > United States > New York (0.04)
- North America > Jamaica (0.04)
- (4 more...)
- Research Report > Experimental Study (1.00)
- Research Report > New Finding (0.67)
- Banking & Finance (0.67)
- Information Technology > Services (0.46)
Learning from Both Structural and Textual Knowledge for Inductive Knowledge Graph Completion
In this paper, we propose a two-stage framework that imposes both structural and textual knowledge to learn rule-based systems. In the first stage, we compute a set of triples with confidence scores (called soft triples) from a text corpus by distant supervision, where a textual entailment model with multi-instance learning is exploited to estimate whether a given triple is entailed by a set of sentences. In the second stage, these soft triples are used to learn a rule-based model for KGC.
- Asia > China > Guangdong Province > Guangzhou (0.04)
- Asia > China > Guangdong Province > Shenzhen (0.04)
- Oceania > Australia > Victoria > Melbourne (0.04)
- North America > United States > Illinois > Champaign County > Champaign (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- (2 more...)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Temporal Reasoning (0.74)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Semantic Networks (0.48)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (0.46)
Duality-Induced Regularizer for Tensor Factorization Based Knowledge Graph Completion
Tensor factorization based models have shown great power in knowledge graph completion (KGC). However, their performance usually suffers from the overfitting problem seriously. This motivates various regularizers---such as the squared Frobenius norm and tensor nuclear norm regulariers---while the limited applicability significantly limits their practical usage. To address this challenge, we propose a novel regularizer---namely, \textbf{DU}ality-induced \textbf{R}egul\textbf{A}rizer (DURA)---which is not only effective in improving the performance of existing models but widely applicable to various methods. The major novelty of DURA is based on the observation that, for an existing tensor factorization based KGC model (\textit{primal}), there is often another distance based KGC model (\textit{dual}) closely associated with it.
ReFactorGNNs
In this section, we prove Theorem 1, which we restate here for convenience. Note that the component " " (highlighted in red) in Equation (18) is a sum " (highlighted in blue) is a term that contains State-of-the-art FMs are often trained with training strategies adapted for each model category. In general, we can interpret any auxiliary variable introduced by the optimizer (e.g. the velocity) as However, the specific equations would depend on the optimizer's dynamics and would be hard to The two main design choices in Theorem A.1 are 1) the score function In the paper, we chose DistMult and GD because of their mathematical simplicity, leading to easier-to-read formulas. In this paper, we describe the results on FB15K237_v1_ind under some random seed. One implementation for such evaluation can be found in GraIL's codebase.
DANS-KGC: Diffusion Based Adaptive Negative Sampling for Knowledge Graph Completion
Negative sampling (NS) strategies play a crucial role in knowledge graph representation. In order to overcome the limitations of existing negative sampling strategies, such as vulnerability to false negatives, limited generalization, and lack of control over sample hardness, we propose DANS-KGC (Diffusion-based Adaptive Negative Sampling for Knowledge Graph Completion). DANS-KGC comprises three key components: the Difficulty Assessment Module (DAM), the Adaptive Negative Sampling Module (ANS), and the Dynamic Training Mechanism (DTM). DAM evaluates the learning difficulty of entities by integrating semantic and structural features. Based on this assessment, ANS employs a conditional diffusion model with difficulty-aware noise scheduling, leveraging semantic and neighborhood information during the denoising phase to generate negative samples of diverse hardness. DTM further enhances learning by dynamically adjusting the hardness distribution of negative samples throughout training, enabling a curriculum-style progression from easy to hard examples. Extensive experiments on six benchmark datasets demonstrate the effectiveness and generalization ability of DANS-KGC, with the method achieving state-of-the-art results on all three evaluation metrics for the UMLS and Y AGO3-10 datasets.
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (0.93)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Expert Systems (0.93)
- (2 more...)
DrKGC: Dynamic Subgraph Retrieval-Augmented LLMs for Knowledge Graph Completion across General and Biomedical Domains
Xiao, Yongkang, Zhang, Sinian, Dai, Yi, Zhou, Huixue, Hou, Jue, Ding, Jie, Zhang, Rui
Knowledge graph completion (KGC) aims to predict missing triples in knowledge graphs (KGs) by leveraging existing triples and textual information. Recently, generative large language models (LLMs) have been increasingly employed for graph tasks. However, current approaches typically encode graph context in textual form, which fails to fully exploit the potential of LLMs for perceiving and reasoning about graph structures. To address this limitation, we propose DrKGC (Dynamic Subgraph Retrieval-Augmented LLMs for Knowledge Graph Completion). DrKGC employs a flexible lightweight model training strategy to learn structural embeddings and logical rules within the KG. It then leverages a novel bottom-up graph retrieval method to extract a subgraph for each query guided by the learned rules. Finally, a graph convolutional network (GCN) adapter uses the retrieved subgraph to enhance the structural embeddings, which are then integrated into the prompt for effective LLM fine-tuning. Experimental results on two general domain benchmark datasets and two biomedical datasets demonstrate the superior performance of DrKGC. Furthermore, a realistic case study in the biomedical domain highlights its interpretability and practical utility.
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.28)
- North America > United States > Michigan > Washtenaw County > Ann Arbor (0.14)
- North America > Canada > Ontario > Toronto (0.04)
- (2 more...)